-
Notifications
You must be signed in to change notification settings - Fork 2.4k
Update cerebras.ts #9024
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Update cerebras.ts #9024
Conversation
Updated maxTokens, contextWindow and description.
| "zai-glm-4.6": { | ||
| maxTokens: 16_384, | ||
| contextWindow: 128000, | ||
| maxTokens: 40960, |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Hey @sebastiand-cerebras, I'm up for whatever you all think is best but personally I think reserving that many tokens for output is a little much for a 131k context window.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Hi Matt,
Thank you so much for your message! We decided to go for this value across our external messaging but I see that it's rather a theoretical value. Would you want me to reduce it in the case of roocode? Wouldn't be an issue!
Btw, I am a big fan :)
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thank you! Yeah personally I've found 16k to be enough for coding, and then you reserve ~115k for task history before condensing etc.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I updated the maxTokens. What do you think?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks good, thanks!
Description
Updated maxTokens, contextWindow and description.
Test Procedure
n/a
Pre-Submission Checklist
Documentation Updates
Additional Notes
Thank you so much!
Important
Update
zai-glm-4.6model incerebras.tswith increasedmaxTokens,contextWindow, and reviseddescription.cerebras.ts, updatedzai-glm-4.6model:maxTokensincreased from 16,384 to 40,960.contextWindowincreased from 128,000 to 131,072.descriptionupdated to "Highly intelligent general purpose model with up to 1,000 tokens/s".This description was created by
for 6260be2. You can customize this summary. It will automatically update as commits are pushed.